Goto

Collaborating Authors

 Pittsfield


RELIC: Investigating Large Language Model Responses using Self-Consistency

Cheng, Furui, Zouhar, Vilém, Arora, Simran, Sachan, Mrinmaya, Strobelt, Hendrik, El-Assady, Mennatallah

arXiv.org Artificial Intelligence

Large Language Models (LLMs) are notorious for blending fact with fiction and generating non-factual content, known as hallucinations. To tackle this challenge, we propose an interactive system that helps users obtain insights into the reliability of the generated text. Our approach is based on the idea that the self-consistency of multiple samples generated by the same LLM relates to its confidence in individual claims in the generated texts. Using this idea, we design RELIC, an interactive system that enables users to investigate and verify semantic-level variations in multiple long-form responses. This allows users to recognize potentially inaccurate information in the generated text and make necessary corrections. From a user study with ten participants, we demonstrate that our approach helps users better verify the reliability of the generated text. We further summarize the design implications and lessons learned from this research for inspiring future studies on reliable human-LLM interactions.


Stock Forecast Based On a Predictive Algorithm

#artificialintelligence

This forecast is part of the Dividends Package, as one of I Know First's quantitative investment solutions. We determine the best stocks carrying a dividend by screening our database daily using our advanced algorithm. Package Name: Dividend Stocks Forecast Recommended Positions: Long Forecast Length: 1 Year (2/7/21 – 2/7/22) I Know First Average: 41.1% For this 1 Year forecast the algorithm had successfully predicted 10 out of 10 movements. The highest trade return came from BHLB, at 61.06%. The suggested trades for CMA and STLD also had notable 1 Year yields of 55.88% and 55.01%, respectively.


Aggregating Content and Network Information to Curate Twitter User Lists

Greene, Derek, Sheridan, Gavin, Smyth, Barry, Cunningham, Pádraig

arXiv.org Artificial Intelligence

Twitter introduced user lists in late 2009, allowing users to be grouped according to meaningful topics or themes. Lists have since been adopted by media outlets as a means of organising content around news stories. Thus the curation of these lists is important - they should contain the key information gatekeepers and present a balanced perspective on a story. Here we address this list curation process from a recommender systems perspective. We propose a variety of criteria for generating user list recommendations, based on content analysis, network analysis, and the "crowdsourcing" of existing user lists. We demonstrate that these types of criteria are often only successful for datasets with certain characteristics. To resolve this issue, we propose the aggregation of these different "views" of a news story on Twitter to produce more accurate user recommendations to support the curation process.


Tenth Anniversary of the Plastics Color Formulation Tool

Cheetham, William E.

AI Magazine

Since 1994, GE Plastics has employed a case-based reasoning (CBR) tool that determines color formulas that match requested colors. This tool, called FormTool, has saved GE millions of dollars in productivity and material (that is, colorant) costs. The technology developed in FormTool has been used to create an online color-selection tool for our customers called ColorXpress Select. A customer innovation center has been developed around the FormTool software.